虽然基于深度学习的方法表明了皮肤病学诊断任务中的专家级表现,但它们还显示出对某些人口统计学属性,尤其是皮肤类型(例如,光对黑暗)的偏见,必须解决公平的关注。我们提出了圆圈,这是一种肤色不变的深度表示学习方法,可改善皮肤病变分类的公平性。通过利用正规化损失来鼓励具有相同诊断的图像但皮肤类型不同以具有相似的潜在表示,对圆圈进行了对图像进行分类的训练。通过广泛的评估和消融研究,我们证明了在跨越6种菲茨帕特里克皮肤类型和114种疾病的16K+图像上评估时,Circle的表现优于最先进的表现,使用分类精度,平等的机会差异(对于光与黑暗组),和归一化精度范围,这是我们提出的一种新措施,以评估多个皮肤类型组的公平性。
translated by 谷歌翻译
深度学习模型在自动化皮肤病变诊断方面取得了巨大成功。但是,在这些模型的预测中,种族差异通常不足以说明深色皮肤类型的病变,并且诊断准确性较低,因此受到很少的关注。在本文中,我们提出了Fairdisco,这是一个带有对比度学习的解开深度学习框架,它利用一个额外的网络分支来消除敏感属性,即从表示的表现形式中的皮肤型信息和另一个对比分支来增强特征提取。我们将Fairdisco与三种公平方法进行了比较,即重新采样,重新加权和属性 - 在两个新发布的具有不同皮肤类型的皮肤病变数据集上:Fitzpatrick17k和多样的皮肤病学图像(DDI)。我们为多个类别和敏感属性任务调整了两个基于公平的指标DPM和EOM,突出了皮肤病变分类中的皮肤型偏差。广泛的实验评估证明了Fairdisco的有效性,对皮肤病变分类任务的表现更公平,更出色。
translated by 谷歌翻译
现有的可解释人工智能(XAI)算法的界限仅限于技术用户对解释性的需求所基于的问题。这项研究范式不成比例地忽略了XAI的非技术最终用户的较大群体,他们没有技术知识,但需要在其AI-ASS辅助批判性决定中进行解释。缺乏以解释性为重点的功能支持可能会阻碍在医疗保健,刑事司法,金融和自动驾驶系统等高风险领域中对AI的安全和负责任的使用。在这项工作中,我们探讨了如何设计为最终用户的关键任务量身定制的XAI如何激发新技术问题的框架。为了引起用户对XAI算法的解释和要求,我们首先将八个解释表格确定为AI研究人员和最终用户之间的通信工具,例如使用功能,示例或规则来解释。然后,我们在实现不同的解释目标(例如验证AI决策并改善用户的预测结果)的背景下,使用32名外行参与者进行用户研究。基于用户研究结果,我们确定并提出新颖的XAI技术问题,并根据用户的解释目标提出评估度量验证能力。我们的工作表明,在最终用户使用XAI中解决技术问题可以激发新的研究问题。这样的最终用户启发的研究问题有可能通过使人工智能民主化并确保在关键领域中对AI负责使用,从而促进社会利益。
translated by 谷歌翻译
医学成像是现代医学治疗和诊断的基石。但是,对于特定静脉局体任务的成像方式的选择通常涉及使用特定模式的可行性(例如,短期等待时间,低成本,快速获取,辐射/侵入性降低)与临床上的预期性能之间的权衡。任务(例如,诊断准确性,治疗计划的功效和指导)。在这项工作中,我们旨在运用从较不可行但表现更好(优越)模式中学到的知识,以指导利用更可行但表现不佳(劣等)模式,并将其转向提高性能。我们专注于深度学习用于基于图像的诊断。我们开发了一个轻量级的指导模型,该模型在训练仅消耗劣质模式的模型时利用从优越方式中学到的潜在表示。我们在两种临床应用中检查了我们方法的优势:从临床和皮肤镜图像中的多任务皮肤病变分类以及来自多序列磁共振成像(MRI)和组织病理学图像的脑肿瘤分类。对于这两种情况,我们在不需要出色的模态的情况下显示出劣质模式的诊断性能。此外,在脑肿瘤分类的情况下,我们的方法的表现优于在上级模态上训练的模型,同时产生与推理过程中使用两种模态的模型相当的结果。
translated by 谷歌翻译
Explainable artificial intelligence (XAI) is essential for enabling clinical users to get informed decision support from AI and comply with evidence-based medical practice. Applying XAI in clinical settings requires proper evaluation criteria to ensure the explanation technique is both technically sound and clinically useful, but specific support is lacking to achieve this goal. To bridge the research gap, we propose the Clinical XAI Guidelines that consist of five criteria a clinical XAI needs to be optimized for. The guidelines recommend choosing an explanation form based on Guideline 1 (G1) Understandability and G2 Clinical relevance. For the chosen explanation form, its specific XAI technique should be optimized for G3 Truthfulness, G4 Informative plausibility, and G5 Computational efficiency. Following the guidelines, we conducted a systematic evaluation on a novel problem of multi-modal medical image explanation with two clinical tasks, and proposed new evaluation metrics accordingly. Sixteen commonly-used heatmap XAI techniques were evaluated and found to be insufficient for clinical use due to their failure in G3 and G4. Our evaluation demonstrated the use of Clinical XAI Guidelines to support the design and evaluation of clinically viable XAI.
translated by 谷歌翻译
药物重新利用可以加速鉴定有效化合物用于针对SARS-COV-2的临床使用,并具有先前存在的临床安全数据和已建立的供应链的优势。 RNA病毒(例如SARS-COV-2)操纵细胞途径并诱导亚细胞结构的重组以支持其生命周期。可以使用生物成像技术来量化这些形态学的变化。在这项工作中,我们开发了DEEMD:使用深层神经网络模型在多个实例学习框架内的计算管道,以基于对公开可用RXRX19A数据集的形态分析来确定针对SARS-COV-2有效的推定治疗方法。该数据集由SARS-COV-2未感染的细胞和受感染细胞的荧光显微镜图像组成,有或没有药物治疗。 Deemd首先提取歧视性形态学特征,以产生来自未感染和感染细胞的细胞形态特征。然后在统计模型中使用这些形态学特征,以根据与未感染细胞的相似性估算受感染细胞的应用治疗疗效。 DEEMD能够通过弱监督定位受感染的细胞,而无需任何昂贵的像素级注释。 DEEMD确定已知的SARS-COV-2抑制剂,例如Remdesivir和Aloxistatin,支持我们方法的有效性。可以在其他新兴病毒和数据集上探索DEEMD,以便将来快速识别候选抗病毒药治疗}。我们的实施可在线网络https://www.github.com/sadegh-saberian/deemd
translated by 谷歌翻译
我们提出了一种自动化方法来检测3D全身皮肤表面扫描的纵向轨道皮肤病。所获取的对象的3D网格未打开到2D纹理图像,其中训练的对象检测模型更快R-CNN,本地化了2D域内的病变。这些检测到的皮肤病变被映射到受试者的3D表面,并且对于多次成像的受试者,我们将基于图形的匹配过程构建到纵向跟踪病变,其考虑了对网眼对和对应病变的测距率的解剖对应关系和病变间测地距距离。我们使用3DBodyTex评估了所提出的方法,该方法是由3D扫描成像200人类受试者的彩色皮肤(纹理网格)组成的公开数据集。我们手动注释出现在人眼中以含有着色皮肤病变以及跟踪在不同姿势成像的同一主题上发生的病变子集的位置。我们的结果与三个人类注释者相比,建议训练的更快的R-CNN检测与人类注释器类似的性能水平的病变。我们的病变跟踪算法在一组检测到的对象对象的突出病变对不同姿势上突出的突出病变的平均匹配精度,以及71%的平均纵向精度,当由于病变检测而包括额外的误差。由于目前没有其他3D全身皮肤病患者的大规模公共可用数据集,我们公开发布超过25,000个3DBodytex手动注释,我们希望进一步研究全身皮肤病因分析。
translated by 谷歌翻译
Deep learning can extract rich data representations if provided sufficient quantities of labeled training data. For many tasks however, annotating data has significant costs in terms of time and money, owing to the high standards of subject matter expertise required, for example in medical and geophysical image interpretation tasks. Active Learning can identify the most informative training examples for the interpreter to train, leading to higher efficiency. We propose an Active learning method based on jointly learning representations for supervised and unsupervised tasks. The learned manifold structure is later utilized to identify informative training samples most dissimilar from the learned manifold from the error profiles on the unsupervised task. We verify the efficiency of the proposed method on a seismic facies segmentation dataset from the Netherlands F3 block survey, significantly outperforming contemporary methods to achieve the highest mean Intersection-Over-Union value of 0.773.
translated by 谷歌翻译
Hydrocarbon prospect risking is a critical application in geophysics predicting well outcomes from a variety of data including geological, geophysical, and other information modalities. Traditional routines require interpreters to go through a long process to arrive at the probability of success of specific outcomes. AI has the capability to automate the process but its adoption has been limited thus far owing to a lack of transparency in the way complicated, black box models generate decisions. We demonstrate how LIME -- a model-agnostic explanation technique -- can be used to inject trust in model decisions by uncovering the model's reasoning process for individual predictions. It generates these explanations by fitting interpretable models in the local neighborhood of specific datapoints being queried. On a dataset of well outcomes and corresponding geophysical attribute data, we show how LIME can induce trust in model's decisions by revealing the decision-making process to be aligned to domain knowledge. Further, it has the potential to debug mispredictions made due to anomalous patterns in the data or faulty training datasets.
translated by 谷歌翻译
Humans exhibit disagreement during data labeling. We term this disagreement as human label uncertainty. In this work, we study the ramifications of human label uncertainty (HLU). Our evaluation of existing uncertainty estimation algorithms, with the presence of HLU, indicates the limitations of existing uncertainty metrics and algorithms themselves in response to HLU. Meanwhile, we observe undue effects in predictive uncertainty and generalizability. To mitigate the undue effects, we introduce a novel natural scene statistics (NSS) based label dilution training scheme without requiring massive human labels. Specifically, we first select a subset of samples with low perceptual quality ranked by statistical regularities of images. We then assign separate labels to each sample in this subset to obtain a training set with diluted labels. Our experiments and analysis demonstrate that training with NSS-based label dilution alleviates the undue effects caused by HLU.
translated by 谷歌翻译